Original Paper
Abstract
Background: The exponential growth of digital information has led to an unprecedented expansion in the volume of unstructured text data. Efficient classification of these data is critical for timely evidence synthesis and informed decision-making in health care. Machine learning techniques have shown considerable promise for text classification tasks. However, multiclass classification of papers by study publication type has been largely overlooked compared to binary or multilabel classification. Addressing this gap could significantly enhance knowledge translation workflows and support systematic review processes.
Objective: This study aimed to fine-tune and evaluate domain-specific transformer-based language models on a gold-standard dataset for multiclass classification of clinical literature into mutually exclusive categories: original studies, reviews, evidence-based guidelines, and nonexperimental studies.
Methods: The titles and abstracts of McMaster’s Premium Literature Service (PLUS) dataset comprising 162,380 papers were used for fine-tuning seven domain-specific transformers. Clinical experts classified the papers into four mutually exclusive publication types. PLUS data were split in an 80:10:10 ratio into training, validation, and testing sets, with the Clinical Hedges dataset used for external validation. A grid search evaluated the impact of class weight (CW) adjustments, learning rate (LR), batch size (BS), warmup ratio, and weight decay (WD), totaling 1890 configurations. Models were assessed using 10 metrics, including the area under the receiver operating characteristic curve (AUROC), the F1-score (harmonic mean of precision and recall), and Matthew’s correlation coefficient (MCC). The performance of individual classes was assessed using a one-to-rest approach, and overall performance was assessed using the macro average. Optimal models identified from validation results were further tested on both PLUS and Clinical Hedges, with calibration assessed visually.
Results: Ten best-performing models achieved macro AUROC≥0.99, F1-score≥0.89, and MCC≥0.88 on the validation and testing sets. Performance declined on Clinical Hedges. Models were consistently better at classifying original studies and reviews. Biomedical Bidirectional Encoder Representations from Transformers (fine-tuned on biomedical text; BioBERT)–based models had superior calibration performance, especially for original studies and reviews. Optimal configurations for search included lower LRs (1 × 10–5 and 3 × 10–5), midrange BSs (32–128), and lower WD (0.005-0.010). CW adjustments improved recall but generally reduced performance on other metrics. Models generally struggled with accurately classifying nonexperimental and guideline studies, potentially due to class imbalance and content heterogeneity.
Conclusions: This study used a comprehensive hyperparameter search to highlight the effectiveness of fine-tuned transformer models, notably BioBERT variants, for multiclass clinical literature classification. Although class weighting generally decreased overall performance, addressing class imbalance through alternative methods, such as hierarchical classification or targeted resampling, warrants future exploration. Hyperparameter configurations were crucial for robust performance, aligning with the previous literature. These findings support future modeling research and practical deployment in human-in-the-loop systems to support knowledge synthesis and translation workflows with the findings from this work.
doi:10.2196/77311
Keywords
Introduction
Background
The exponential growth of health evidence has led to an unprecedented expansion in the volume of unstructured text data. For instance, PubMed, a leading repository of biomedical literature, has over 36 million papers indexed as of 2025, with approximately 1 million new papers added annually [,]. This highlights the critical need for automated methods to classify, organize, and retrieve relevant information efficiently.
Machine Learning for Text Classification Tasks
Text classification is a key natural language processing (NLP) task that involves assigning predefined categories to unstructured text [,]. It underpins various applications, including sentiment analysis and spam detection [,]. Deep learning, compared to rule-based [-] or shallow [,-] learning approaches, has significantly advanced text classification by automating feature extraction and improving contextual understanding []. Recurrent neural networks (RNNs), particularly their bidirectional and gated variants, demonstrate better performance by capturing sequential relationships in text [-]. However, they suffer from computational inefficiencies, vanishing gradients, and limitations in processing long-range dependencies [-].
Transformer-based models, such as Bidirectional Encoder Representations from Transformers (BERT), have overcome these challenges by introducing self-attention mechanisms that consider all tokens in parallel, improving both efficiency and contextual embedding [-]. Pretrained models, including domain-specific variants such as Biomedical Bidirectional Encoder Representations from Transformers (fine-tuned on biomedical text; BioBERT) [], Scientific Bidirectional Encoder Representations from Transformers (SciBERT) [], and Biomedical Document Link Bidirectional Encoder Representations from Transformers (BioLinkBERT) [], leverage large-scale biomedical corpora to enhance classification performance in specialized fields. Transfer learning further optimizes these models by adapting them to new tasks with minimal labeled data, making them highly effective for text classification [].
Medical Literature Classification
Traditional indexing methods, such as Medical Subject Headings (MeSH), improve literature retrieval but suffer from delays (often taking months for new papers to be indexed) and inconsistencies due to subjective keyword selection [,]. Machine learning (ML)–based approaches have been explored to classify medical literature into binary, multilabel, and multiclass categories. Early rule-based systems, such as the Medical Text Indexer (MTI), automated MeSH indexing, but newer neural network models, such as MTI-NeXt, significantly improved recall and efficiency [,]. ML has also been used to classify papers based on study type, topic, or methodological rigor, supporting systematic reviews and clinical decision-making [-].
Binary classifiers have been used to assess methodological soundness or relevance to systematic reviews [-]. However, multiclass classification, where papers must be assigned a single category among multiple mutually exclusive classes, presents additional challenges [-]. Traditional approaches decompose the problem into multiple binary tasks using one-versus-one (OvO) or one-versus-rest (OvR) methods, but these are computationally expensive or, for OvR, yield ambiguous or poorly calibrated predictions when multiple classifiers fire [-]. In contrast, transformers natively support multiclass classification through multiheaded SoftMax activation []. Yet, limited studies have examined their effectiveness in the multiclass classification of clinical literature, and existing work has not fully addressed the impact of pretraining, class imbalance, or optimal hyperparameter selection [,].
Objective
The objective of this research was to fine-tune and evaluate the performance of seven domain-specific, encoder-only transformers using various hyperparameter configurations on the task of multiclass classification of published clinical literature into original studies, reviews, evidence-based guidelines, and nonexperimental studies. We leveraged two datasets curated and annotated by the Health Information Research Unit (HIRU) at McMaster University, with the goal of identifying a multiclass model to support evidence processing early after publication and before indexing tasks are completed.
Methods
Data Source and Preprocessing
HIRU is a pioneer in providing curated evidence services targeted to clinicians worldwide. Originally, HIRU curated the Clinical Hedges dataset in 2000, which comprises classifications and critical appraisal of ~49,000 papers across 161 journals indexed in MEDLINE []. Each paper was manually classified by domain experts into mutually exclusive study formats, and clinically relevant papers that reported the findings of an original study or a review were subsequently labeled for research purposes and assessed for methodological rigor.
Subsequently, McMaster’s Premium Literature Service (PLUS) dataset, which classifies and appraises clinical research reports indexed in PubMed, was initiated in 2003. Through the ongoing PLUS process, HIRU has continued to grow a database of clinically relevant papers appraised at the time of publication using Clinical Hedges’s criteria and classifications with some modifications. Specifically, PLUS involves automated daily searches of PubMed using a sensitive methods filter adapted from Clinical Queries and applied to ~125 journal titles (with some expansion during the COVID-19 pandemic to all journals and using a COVID-19–based filter) []. To date, the PLUS dataset includes over 150,000 clinical papers classified by publication type into mutually exclusive classes: (1) original studies, (2) reviews, (3) evidence-based guidelines, or (4) nonexperimental studies (for indexed items such as case studies, general or philosophical discussions of a topic without original observations and without a statement that the purpose was to review or appraise a body of knowledge, secondary publications, letters, or commentaries) [].
All papers that were retrieved and assessed through PLUS from inception to 2023 were included. The dataset was randomly split in an 80:10:10 ratio into training, validation, and testing sets, and these subsets were termed PLUS-train, PLUS-validate, and PLUS-test, respectively. The original Clinical Hedges dataset was used for external testing. The titles and abstracts of papers were combined and tokenized using the pretrained model’s corresponding tokenizer and were used as inputs. Inputs longer than the maximum token length of 512 were truncated, and shorter inputs were padded to ensure uniform input length. No text normalization was performed.
Model Configurations and Hyperparameters
We selected seven pretrained models to fine-tune, based on their performance in the previous literature and the Biomedical Language Understanding and Reasoning Benchmark (BLURB) leaderboard [-,-]. Each model was trained with or without class weight (CW) adjustments. The training process minimized cross-entropy loss. We used a linear learning rate (LR) scheduler. The AdamW optimizer [] was used with the default β1 of 0.9 and β2 of 0.999. We enabled mixed precision training for faster training and less memory usage. In the case of insufficient memory for a particular batch size (BS), we enabled gradient accumulation. We fine-tuned for at most 10 epochs with an early stopping patience of 3, where training was prematurely terminated if the cross-entropy loss on the PLUS-validate set failed to improve for 3 consecutive epochs. The weights from the epoch with the lowest loss on the PLUS-validate set were selected to mitigate overfitting. A grid search of the LR, BS, warmup ratio (WR), and weight decay (WD) was conducted. Overall, models were trained with 1890 configurations (). Each model’s output layer produced four logits, one for each class, which were converted to probabilities using the SoftMax function. During fine-tuning, the models minimized categorical cross-entropy loss.
| Parameter | Count | Values |
| Pretrained model | 7 | BioBERTa [], BioELECTRAb [], BioLinkBERTc [], BiomedBERTd (abstracts only) [], BiomedBERT (abstracts+full text) [], SciBERTe-cased [], SciBERT-uncased [] |
| CWf adjustment | 2 | Yes, no |
| LRg | 3 | 1 × 10–5, 3 × 10–5, 5 × 10–5 |
| BSh | 5 | 16, 32, 64, 128, 256 |
| WRi | 3 | 0.05, 0.10, 0.20 |
| WDj | 3 | 0.005, 0.010, 0.015 |
aBioBERT: Biomedical Bidirectional Encoder Representations from Transformers (fine-tuned on biomedical text).
bBioELECTRA: Biomedical Efficiently Learning an Encoder that Classifies Token Replacements Accurately.
cBioLinkBERT: Biomedical Document Link Bidirectional Encoder Representations from Transformers.
dBiomedBERT: Biomedical Bidirectional Encoder Representations from Transformers (trained entirely on biomedical text). Formerly known as PubMedBERT.
eSciBERT: Scientific Bidirectional Encoder Representations from Transformers.
fCW: class weight.
gLR: learning rate.
hBS: batch size.
iWR: warmup ratio.
jWD: weight decay.
Class Weight Calculation
CWs were calculated using the following formula:
where N is the total number of samples and ni is the number of samples in class “i.”
Model Evaluation
Macrolevel Performance of Models
We presented model results using the macro average of cross-entropy loss, the Brier score, the area under the receiver operating characteristic curve (AUROC), the average precision (AP), recall, precision, accuracy, F1-scores (harmonic mean of precision and recall) and F2-scores (harmonic mean of precision and recall, with an emphasis on recall), and Matthew’s correlation coefficient (MCC). The interpretation of these metrics is tabulated in Table S1 in . The macro average was selected to mitigate bias toward more prevalent classes.
We grouped models based on the values of the six configurations in to assess how model performance was affected. Performance differences on the validation set between model configurations were presented using means and 95% CIs and visualized using bar plots.
Selecting and Testing the Best Models
We narrowed the models to those that achieved the best macro average performance on one or more evaluation metrics on the validation set. These models were then further evaluated on the testing set and Clinical Hedges []. For individual classes, we used an OvR approach, where each class was treated as a separate binary classification problem. Specifically, for each class, we labeled instances of that class as the positive class and instances of all other classes as the negative class. Metrics on evidence-based guidelines were not considered for the Clinical Hedges test as the dataset did not contain this class. We used AUROC, the F1-score, and the MCC as the primary evaluation metrics. We estimated 95% CIs using bootstrapping (ie, repeated sampling with replacement) over 1000 iterations []. Confusion matrices were provided to examine classification errors between classes. Calibration plots of each model were presented and visually inspected. Points below the diagonal indicated that the true proportion of the class is lower than the average predicted probability, meaning the model is overconfident for that class. Points above the diagonal indicated underconfidence. A perfect diagonal line meant the model is perfectly calibrated, reflecting reliable performance across the full probability range.
Hardware and Software
All fine-tuning was performed on the Cedar cluster provided by the Digital Research Alliance of Canada. Each model was trained on a single NVIDIA V100 Volta graphics processing unit (GPU; 32 GB memory), with access to eight central processing unit cores and 40 GB of memory. Details of the software environment are listed in Table S2 in . All software development was carried out using Visual Studio Code and Python 3.11.5. Pretrained models were obtained using the Hugging Face Transformers library, while evaluation tasks were performed with PyTorch. Data management and statistical analysis were conducted with Pandas, NumPy, and scikit-learn, and matplotlib was used for data visualization.
Ethical Considerations
This study involved the processing of only published biomedical and clinical literature, so ethics approval was not required.
Results
Characteristics of Datasets
The training of all 1890 models took 5162.42 GPU-hours. Each model used an average of 2.73 (SD 0.52) hours to train. A total of 162,380 records from PLUS were used in this study, of which 129,904 (80%) were used for training (). The class distribution of the training, validation, and testing sets was similar, in which original studies, reviews, evidence-based guidelines, and nonexperimental studies accounted for approximately 65.0% (n=84,398), 24.4% (n=31,684), 1.8% (n=2318), and 8.9% (n=11,504) of the dataset, respectively. Clinical Hedges included a total of 48,044 papers, of which most were original (n=25,747, 53.6%) or nonexperimental (n=19,234, 40.0%) studies. There were no evidence-based guidelines in Clinical Hedges.
| Dataset | Study type | Total (N=162,380), n (%) | ||||
| Original studies, n (%) | Reviews, n (%) | Evidence-based guidelines, n (%) | Nonexperimental studiesa, n (%) | |||
| PLUSb-train | 84,398 (65.0) | 31,684 (24.4) | 2318 (1.8) | 11,504 (8.9) | 129,904 (80.0) | |
| PLUS-validate | 10,640 (65.5) | 3898 (24.0) | 293 (1.8) | 1407 (8.7) | 16,238 (10.0) | |
| PLUS-test | 10,496 (64.6) | 3989 (24.6) | 318 (2.0) | 1435 (8.8) | 16,238 (10.0) | |
| PLUS total | 105,534 (65.0) | 39,571 (24.4) | 2929 (1.8) | 14,346 (8.8) | 162,380 (100.0) | |
| Clinical Hedges | 25,747 (53.6) | 3063 (6.4) | N/Ac | 19,234 (40.0) | 48,044 (29.6) | |
aCase studies, general or philosophical discussions without original observations or a clear purpose to review or appraise a body of knowledge, secondary publications, letters, or commentaries.
bPLUS: Premium Literature Service.
cN/A: not applicable. The Clinical Hedges dataset includes papers from 2000; guidelines were not as prevalent at the time, and the label was not used.
Aggregated Performance on PLUS-Validate
The metric macro averages of the different model configurations on the PLUS-validate set are summarized in , Figures S1-S6 in , and Table S1 in . Models without CW adjustments showed better performance across all metrics except for recall and the F2-score (Figure S2 in ). Models with lower LRs generally performed better than those with higher LRs (Figure S3 in ). A BS of 16 resulted in worse performance, and a BS of 256 had wider variance; the other sizes showed relatively mixed performance (Figure S4 in ). There was no clear trend among the WRs (Figure S5 in ), and a smaller WD marginally improved performance (Figure S6 in ).
| Configuration and models | AUROCb, mean (95% CI) | F1-score, mean (95% CI) | MCCc, mean (95% CI) | |||
| Pretrained model | ||||||
| BioBERTd | 0.994 (0.994-0.994) | 0.897 (0.895-0.898) | 0.884 (0.883-0.886) | |||
| BioELECTRAe | 0.993 (0.993-0.994) | 0.896 (0.895-0.897) | 0.883 (0.882-0.884) | |||
| BioLinkBERTf | 0.994 (0.994-0.994) | 0.898 (0.896-0.899) | 0.885 (0.884-0.886) | |||
| BiomedBERTg (abstract only) | 0.992 (0.990-0.995) | 0.895 (0.891-0.899) | 0.882 (0.876-0.887) | |||
| BiomedBERT (abstract+full text) | 0.994 (0.994-0.994) | 0.897 (0.895-0.898) | 0.884 (0.883-0.886) | |||
| SciBERTh-cased | 0.994 (0.994-0.994) | 0.894 (0.893-0.896) | 0.881 (0.880-0.883) | |||
| SciBERT-uncased | 0.994 (0.994-0.994) | 0.897 (0.896-0.898) | 0.884 (0.883-0.885) | |||
| CWiadjustment | ||||||
| No | 0.995 (0.995-0.995) | 0.903 (0.903-0.904) | 0.891 (0.890-0.891) | |||
| Yes | 0.993 (0.992-0.993) | 0.889 (0.888-0.890) | 0.877 (0.875-0.878) | |||
| LRj | ||||||
| 1 × 10–5 | 0.995 (0.994-0.995) | 0.901 (0.900-0.901) | 0.889 (0.888-0.889) | |||
| 3 × 10–5 | 0.993 (0.992-0.995) | 0.895 (0.894-0.897) | 0.883 (0.880-0.885) | |||
| 5 × 10–5 | 0.993 (0.993-0.993) | 0.892 (0.891-0.893) | 0.879 (0.878-0.880) | |||
| BSk | ||||||
| 16 | 0.992 (0.992-0.993) | 0.895 (0.895-0.896) | 0.882 (0.881-0.883) | |||
| 32 | 0.993 (0.993-0.994) | 0.896 (0.895-0.897) | 0.884 (0.883-0.885) | |||
| 64 | 0.994 (0.994-0.994) | 0.896 (0.895-0.897) | 0.884 (0.882-0.885) | |||
| 128 | 0.995 (0.995-0.995) | 0.897 (0.896-0.898) | 0.885 (0.884-0.886) | |||
| 256 | 0.994 (0.992-0.996) | 0.896 (0.893-0.899) | 0.884 (0.880-0.888) | |||
| WRl | ||||||
| 0.05 | 0.994 (0.994-0.994) | 0.895 (0.895-0.896) | 0.883 (0.882-0.884) | |||
| 0.10 | 0.994 (0.994-0.994) | 0.896 (0.895-0.897) | 0.883 (0.883-0.884) | |||
| 0.20 | 0.993 (0.992-0.995) | 0.897 (0.895-0.899) | 0.884 (0.882-0.887) | |||
| WDm | ||||||
| 0.005 | 0.994 (0.994-0.994) | 0.896 (0.895-0.897) | 0.884 (0.883-0.885) | |||
| 0.010 | 0.994 (0.994-0.994) | 0.896 (0.895-0.897) | 0.884 (0.883-0.885) | |||
| 0.015 | 0.993 (0.992-0.995) | 0.896 (0.894-0.898) | 0.883 (0.881-0.885) | |||
aPLUS: Premium Literature Service.
bAUROC: area under the receiver operating characteristic curve.
cMCC: Matthew’s correlation coefficient.
dBioBERT: Biomedical Bidirectional Encoder Representations from Transformers (fine-tuned on biomedical text).
eBioELECTRA: Biomedical Efficiently Learning an Encoder that Classifies Token Replacements Accurately.
fBioLinkBERT: Biomedical Document Link Bidirectional Encoder Representations from Transformers.
gBiomedBERT: Biomedical Bidirectional Encoder Representations from Transformers (trained entirely on biomedical text). Formerly known as PubMedBERT.
hSciBERT: Scientific Bidirectional Encoder Representations from Transformers.
iCW: class weight.
jLR: learning rate.
kBS: batch size.
lWR: warmup ratio.
mWD: weight decay.
Best-Performing Models
No model had the best macro average in more than one metric on the validation set. The configurations of the 10 best models can be found in Table S3 in . BioBERT (n=4, 40%) and BiomedBERT (n=4, 40%) were the most frequently used pretrained architectures, followed by BioLinkBERT (n=1, 10%) and SciBERT-uncased (n=1, 10%). Most models did not use class weighting (n=8, 80%), and the most common LRs were 1 × 10–5 (n=4, 40%) and 3 × 10–5 (n=4, 40%). The majority of models had a weight decay of either 0.010 (n=3, 30%) or 0.015 (n=5, 50%). For the BS, 2 (20%), 2 (20%), 2 (20%), and 4 (40%)models used a value of 16, 64, 128, and 256, respectively. For the WR, 3 (30%), 3 (30%), and 4 (40%)models used a value of 0.05, 0.10, and 0.20, respectively.
and Table S2 in present the performance of models that achieved the best macro average in AUROC (BiomedBERT: CW=no; LR=3 × 10–5; BS=256; WR=0.20; WD=0.005), the F1-score (BioBERT: CW=no; LR=5 × 10–5; BS=64; WR=0.10; WD=0.015), and the MCC (BiomedBERT: CW=no; LR=1 × 10–5; BS=16; WR=0.20; WD=0.005) on the PLUS-validate set. On the PLUS subsets, the macro average AUROC, F1-score, and MCC were 0.993-0.996, 0.904-0.914, and 0.890-0.902, respectively. Using an OvR approach, the classification performance, from best to worst, was original studies, reviews, nonexperimental studies, and evidence-based guidelines. On Clinical Hedges, macro average performance was significantly lower in AUROC (0.957-0.963), the F1-score (0.817-0.826), and the MCC (0.754-0.765). Class-wise, the models were the best at classifying original studies and the worst at reviews.
The confusion matrices () of the three models indicated that common confusions occurred between nonexperimental studies misclassified as reviews or original studies, and vice versa, on Clinical Hedges. No notable pattern of confusion was present on the PLUS subsets. Upon visual inspection of the calibration plots (), the best-AUROC model (BiomedBERT: CW=no; LR=3 × 10–5; BS=256; WR=0.20; WD=0.005) had the best calibration among the three models. On the PLUS subsets, the models were generally well calibrated, considering that few papers were predicted with probability≥0.10 or ≤0.90, evident by the width of the 95% CI. Models demonstrated poorer calibration on Clinical Hedges, where they were underconfident on original studies and overconfident on reviews and evidence-based guidelines.
The results of all 10 best-performing models across all metrics on PLUS-validate, PLUS-test, and Clinical Hedges can be found in Tables S4, S5, and S6 in , respectively. Confusion matrices and calibration plots for the other seven models can be found in Figures S7-S12 and Figures S13-S19 in , respectively. In general, the best-recall model (SciBERT-uncased: CW=yes; LR=3 × 10–5; BS=256; WR=0.05; WD=0.010) had worse performance than others, and all other models demonstrated similar performance and trends without meaningful differences. The best-loss (BioBERT: CW=no; LR=5 × 10–5; BS=256; WR=0.10; WD=0.015; Figure S13 in ), best-Brier score (BioBERT: CW=no; LR=1 × 10–5; BS=64; WR=0.20; WD=0.015; Figure S14 in ), and best-accuracy (BioBERT: CW=no; LR=1 × 10–5; BS=256; WR=0.05; WD=0.015; Figure S18 in ) models were better calibrated than the others on the PLUS subsets. Calibration on Clinical Hedges was mixed, with all models being underconfident on original studies and overconfident in the other classes, with the best-AP model (BiomedBERT: CW=no; LR=1 × 10–5; BS=128; WR=0.05; WD=0.010; Figure S15 in ) having the best performance based on visual inspection.
| Model (best metric; CWc, LRd, BSe, WRf, WDg) and class | PLUSh-test | Clinical Hedges | ||||||
| AUROC, score (bootstrapped 95% CI) | F1-score, score (bootstrapped 95% CI) | MCC, score (bootstrapped 95% CI) | AUROC, score (bootstrapped 95% CI) | F1-score, score (bootstrapped 95% CI) | MCC, score (bootstrapped 95% CI) | |||
| BiomedBERTi(AUROC; no, 3 × 10–5, 256, 0.20, 0.005) | ||||||||
| Original study | 0.997 (0.996-0.998) | 0.987 (0.986-0.989) | 0.964 (0.960-0.968) | 0.974 (0.973-0.976) | 0.928 (0.925-0.930) | 0.855 (0.850-0.860) | ||
| Review | 0.996 (0.995-0.997) | 0.961 (0.957-0.965) | 0.948 (0.942-0.954) | 0.953 (0.949-0.957) | 0.677 (0.664-0.690) | 0.655 (0.641-0.669) | ||
| Evidence-based guideline | 0.994 (0.991-0.997) | 0.821 (0.784-0.853) | 0.818 (0.780-0.850) | N/Aj | N/A | N/A | ||
| Nonexperimental study | 0.991 (0.988-0.993) | 0.869 (0.856-0.881) | 0.856 (0.841-0.869) | 0.961 (0.959-0.962) | 0.874 (0.871-0.878) | 0.785 (0.780-0.791) | ||
| Macro average | 0.995 (0.993-0.996) | 0.910 (0.899-0.919) | 0.897 (0.885-0.906) | 0.963 (0.961-0.964) | 0.826 (0.821-0.831) | 0.765 (0.759-0.771) | ||
| BioBERTk (F1-score; no, 5 × 10–5, 64, 0.10, 0.015) | ||||||||
| Original study | 0.997 (0.996-0.998) | 0.986 (0.985-0.988) | 0.962 (0.957-0.966) | 0.972 (0.970-0.973) | 0.923 (0.921-0.926) | 0.848 (0.843-0.852) | ||
| Review | 0.996 (0.995-0.997) | 0.962 (0.958-0.966) | 0.950 (0.944-0.955) | 0.948 (0.944-0.952) | 0.655 (0.642-0.669) | 0.633 (0.620-0.648) | ||
| Evidence-based guideline | 0.994 (0.990-0.997) | 0.809 (0.775-0.839) | 0.805 (0.771-0.836) | N/A | N/A | N/A | ||
| Nonexperimental study | 0.992 (0.989-0.993) | 0.872 (0.859-0.883) | 0.859 (0.846-0.872) | 0.958 (0.956-0.960) | 0.871 (0.868-0.875) | 0.781 (0.775-0.786) | ||
| Macro average | 0.995 (0.993-0.996) | 0.907 (0.897-0.917) | 0.894 (0.883-0.904) | 0.959 (0.957-0.961) | 0.817 (0.812-0.822) | 0.754 (0.748-0.760) | ||
| BiomedBERT (MCC; no, 1 × 10–5, 16, 0.20, 0.005) | ||||||||
| Original study | 0.996 (0.995-0.997) | 0.987 (0.986-0.989) | 0.964 (0.959-0.968) | 0.967 (0.965-0.968) | 0.913 (0.911-0.916) | 0.833 (0.828-0.838) | ||
| Review | 0.996 (0.995-0.997) | 0.958 (0.954-0.963) | 0.945 (0.939-0.951) | 0.949 (0.945-0.953) | 0.691 (0.677-0.704) | 0.673 (0.658-0.687) | ||
| Evidence-based guideline | 0.992 (0.987-0.996) | 0.801 (0.763-0.836) | 0.798 (0.761-0.833) | N/A | N/A | N/A | ||
| Nonexperimental study | 0.989 (0.986-0.992) | 0.868 (0.854-0.880) | 0.855 (0.840-0.868) | 0.954 (0.953-0.956) | 0.869 (0.865-0.872) | 0.776 (0.771-0.781) | ||
| Macro average | 0.993 (0.992-0.995) | 0.904 (0.893-0.914) | 0.890 (0.879-0.902) | 0.957 (0.955-0.959) | 0.824 (0.819-0.830) | 0.761 (0.754-0.767) | ||
aAUROC: area under the receiver operating characteristic curve.
bMCC: Matthew’s correlation coefficient.
cCW: class weight.
dLR: learning rate.
eBS: batch size.
fWR: warmup ratio.
gWD: weight decay.
hPLUS: Premium Literature Service.
iBiomedBERT: Biomedical Bidirectional Encoder Representations from Transformers (trained entirely on biomedical text). Formerly known as PubMedBERT.
jN/A: not applicable.
kBioBERT: Biomedical Bidirectional Encoder Representations from Transformers (fine-tuned on biomedical text).


Discussion
Principal Findings
For this study, we leveraged an expert-annotated dataset of over 150,000 papers for tuning state-of-the-art pretrained encoder transformer models using a comprehensive grid search and evaluation process. The findings highlight the effects of hyperparameter settings in text-based multiclass problems. The findings for the case of classifying papers by publication type can inform future research on multiclass classification of clinical text and expected performance when the models are deployed in digital systems to support curating evidence-based clinical literature.
Model Performance
We identified 10 top-performing models and tested these in a set and Clinical Hedges data. These models demonstrated high performance on all study classes with macro AUROC≥0.99, F1-score≥0.89, and MCC≥0.88 on both validation and testing sets. BioBERT-based models tended to have better calibration. These models have the potential to reliably add annotations in the PLUS workflow, especially for original studies and reviews. However, the performance on nonexperimental studies and evidence-based guideline reports was noticeably worse, presumably due to class imbalance and the intrinsic heterogeneity present in these classes of papers []. The performance on Clinical Hedges was also worse, and this could be a result of differences in classification criteria and data drift over time since 2000, as well as changes in the reporting structure of papers [-].
For PLUS, the calibration plots show that most papers are predicted with high confidence (≤10% or ≥90%), and these predictions generally align well with true proportions, indicating reliable classification performance. For Clinical Hedges, the calibration curve for the original class lies mostly above the diagonal, showing that the true proportion of original papers is higher than the predicted probabilities. This suggests the models tend to underestimate their prevalence in this dataset and highlights the importance of monitoring the calibration performance to use the models’ predictions responsibly.
Effect of Model Configurations
All seven pretrained models have the potential to achieve a satisfactory performance for classification in this context when the ideal hyperparameters are combined. However, among the top 10 models, the calibration of the BioBERT models was generally better. For this reason, we believe that BioBERT has the best potential for future examination and deployment for this task.
Typically, the class imbalance should be addressed through resampling or CW adjustments [,]. We found that CW adjustments generally resulted in worse macro average performance, except for recall and those metrics that heavily prioritize recall. Eight of the ten best-performing models did not include CW adjustments, and no notable degradation in performance was seen in evidence-based guidelines and nonexperimental studies compared to the two models that included CW adjustments. Intuitively, with substantial class imbalance, CW adjustment encourages more positive predictions for those classes, increasing both true positives and false positives. Due to drastically lower prevalence of the minority classes, the latter dominated, resulting in a decline in summary metrics despite an increase in recall. This is consistent with the prior literature that found that class imbalance correction techniques may not improve classification performance [-]. Therefore, unless recall for the minority classes is a priority, omitting CW adjustments was preferable in our experiments. Nevertheless, it is apparent that class balance remains an issue after adjustments. Interestingly, in our previous binary classification experiments using Light Gradient Boosting Machine (LightGBM) [] and BERT models [], resampling typically resulted in better performance across most metrics.
The LR is one of the most important hyperparameters in fine-tuning encoder transformers [,]. Our experiment suggests that a lower LR results in better, more consistent performance. This is consistent with the previous literature []. which suggests that higher LRs may not be suitable for finetuning in this context. We recommend that future research on similar tasks include experiments with an LR of 3 × 10–5 or lower for BERT and Efficiently Learning an Encoder that Classifies Token Replacements Accurately (ELECTRA) models. A BS of 256 was the largest we experimented with. We found that a BS of 16 performs much worse than others, and a BS of 256 resulted in more inconsistent performance, as evident through the wider 95% CI. Although 2 of the 10 best-performing models used a BS of 16, their performance was not significantly better than the others. For computational resource efficiency, future experiments should search among BSs of 32 or greater [-].
Regarding the WR, we could not identify any meaningful patterns. A larger ratio seems to improve precision and worsen recall, but the effects are marginal, with a <1% difference on average. Performance on other metrics seems to be mixed as well. Future studies may wish to further explore the effects of the WR in depth. For the WD, a value of 0.005 or 0.010 offered similar performance, with 0.005 being slightly better. A WD of 0.015 should be avoided as the performance was worse and inconsistent.
Practice Implications
Since August 2022, HIRU has implemented a binary BioBERT classifier with the probability threshold set at 99% sensitivity to confidently exclude irrelevant papers for PLUS []. Results indicate that there is a work saving of >60%, and this indicates the strong potential for these models to improve efficiency. This multiclass experiment serves as the next step in the automation of the costly manual appraisal process. For PLUS, the best-performing models would be satisfactory for deployment in human-in-the-loop systems to aid in paper type classification, recognizing this in one step in a series for the human experts. Based on calibration performance and prior implementation experience, a promising strategy would be to automatically classify papers assigned a high probability threshold, such as ≥95% for a given class, while treating lower-confidence predictions as unsure and routing for manual annotation. Providing human reviewers with probability scores can also support their decision-making when classifying papers. New annotations for additional training may also alleviate the performance decrease over time due to data drift [,-].
Regarding other services, Health Evidence, a database focused on appraising and disseminating high-quality public health knowledge synthesis evidence [], has introduced ML into its workflow in a similar fashion []. Specifically, the DistillerSR AI Preview & Rank feature was used, and the minimum probability threshold that would result in perfect recall was identified. Subsequently, the threshold was increased until five false negatives occurred, resulting in doubling of the specificity from 36% to 72%. Over a 3-year implementation period, there was a 70% reduction in the manual screening effort, with an estimated time saving of 382 person-hours. These findings echo the success that HIRU has had and the potential for further efficiency gains from the implementation of the multiclass models for HIRU.
Externally, our models may be of particular interest to systematic reviewers and expert panels who want to quickly obtain relevant studies from among those not yet fully indexed in clinical literature databases. Databases such as PubMed may consider the deployment of such models in similar fashions as HIRU and Health Evidence or in conjunction with existing systems to improve the reliability of paper classifications.
Ultimately, organizations should carefully define the intended role of such models (ie, whether the models are used for initial classification for efficiency gains or deployed to inform human annotators to improve accuracy), as well as establish acceptable trade-offs between error rates, efficiency gains, and cost. Although patient privacy and safety are unlikely to be a concern, copyright issues may be a consideration for full-text use during model training and validation []. It is also crucial to train annotators in the effective interpretation of model outputs and responsible artificial intelligence (AI) use. Finally, our models may not be directly transferable to other databases and systematic review systems with different purposes. Nevertheless, insights from this work regarding model fine-tuning and selection are valuable for future development and benchmarking studies of similar scope.
Comparison With Previous Work
To the best of our knowledge, no other experiment has used a similar corpus for multiclass classification. Therefore, direct comparisons of model performance were not feasible. We instead provide an overview of previous related studies.
Rabby and Berka [] examined various shallow learning (SL) and deep learning (DL) methods to classify COVID-19 papers that had only one label in the LitCovid corpus and included studies in the prevention, treatment, diagnosis, or case report classes. Interestingly, the random forest classifier obtained the best performance, with 0.92 accuracy and macro F1-score using the term frequency–inverse document frequency (TF-IDF) as a feature. BERT, with an LR of 1 × 10–5, an epoch of 5, and a maximum token length of 256, achieved a lower performance of 0.87 accuracy and macro F1-score. We suspect that this could be due to insufficient training data, poor hyperparameter choices, or bias from the lack of a validation set for hyperparameter tuning.
Afzal et al [] used SciBERT and the Unified Medical Language System (UMLS), with bidirectional long short‑term memory (BiLSTM) and active learning, to classify medical notes with the SOAP (Subjective, Objective, Assessment, Plan) structure. Specifically, SciBERT and the UMLS were used to process contextual and semantic information, respectively, before being combined and processed with a dense SoftMax layer. The system was named SOAP-BiomedBERT and achieved an accuracy of 0.98 and an F1-score of 0.99. Compared to SciBERT alone, SOAP-BiomedBERT represents a mild increase of ~0.01 in both accuracy and F1-scores.
Raja et al [] used Bidirectional and Auto-Regressive Transformers (BART) to classify 1000 papers on retinal diseases into 4 classes and 18 labels. BART, as opposed to BERT, is a sequence-to-sequence model that incorporates both an encoder and a decoder. During the training process, inputs are corrupted by masking, deletion, and/or permutation, and the model learns to reconstruct the original text. The architecture makes it more effective at generative and transformation tasks compared to BERT. Classifying for study types (clinical, experimental, or automated), the model achieved an F1-score of 0.92 and an AUROC of 0.91. Although the performance has room for improvement, it was achieved through zero-shot learning without the need for an extensive annotated training set.
Joshi and Abdelfattah [] explored the applicability of six SL multinomial classifiers to predict medical conditions from drug reviews. TF-IDF vectors were used as features, and a random search was used for hyperparameter tuning. The best-performing model, the linear support vector classifier, achieved an F1-score of 0.88. Naive Bayes, in contrast, performed the worst, with a score of 0.64, despite being the fastest model.
Limitations and Future Directions
Several limitations must be noted. First, we were unable to use the full text of papers as inputs due to the intrinsic token limitations of BERT and ELECTRA, although the models performed well on titles and abstracts alone. Experiments using Longformer or other architectures that use the sliding window attention may be warranted. Generative large language models may be worth exploring as well, considering their task-agnostic nature and lower requirement for technical knowledge.
Second, cross-validation and nested cross-validation were not implemented due to computational and storage constraints. The number of papers included in this experiment should effectively reduce the risk of sampling bias. However, studies using a smaller dataset may wish to consider cross-validation during hyperparameter tuning to improve the reliability of the models.
Third, we only attempted CW adjustments to address class imbalance in this work, which did not improve the performance for the minority classes except in terms of recall. We did not explore other methods for comparison, such as other cost-sensitive learning methods; data augmentation, such as synonym replacement or undersampling; and hierarchical classification and ensemble schemes. This is primarily due to resource limitations, considering the large hyperparameter search space we examined. Although the models’ performance on minority classes remained strong, with AUROC≥99% and MCC≥0.80, we nevertheless believe that future studies using a more refined hyperparameter search informed by our results are warranted.
Lastly, all papers used for training were retrieved from clinically focused journals indexed in PubMed. Considering the performance degradation on Clinical Hedges, it is unknown how well the model would generalize to contemporary papers from other commonly used biomedical databases, such as Embase or Scopus, which have broader disciplinary scopes, or from databases with more specific focuses, such as Emcare and PsycINFO, which emphasize allied health, nursing, or behavioral sciences. Differences in subject focus, indexing practices, and paper composition may pose domain-specific challenges. Before external deployment, benchmarking model performance on these sources would be necessary. Applying the model to such databases would also likely require fine-tuning on representative samples from those sources, although this is hindered by the lack of publicly available datasets. Nevertheless, our methods could be readily replicated to achieve similar performance in those contexts.
As future work, the explainability of the models will be explored. One approach is to use integrated gradients or Shapley Additive Explanations (SHAP), which allow for feature attributions to be determined at the token level [-]. This means that the words (or tokens) that are most important to the model classification decision can be identified and shared with human assessors. Alignment of these words with human understanding of the task can improve the transparency and, thus, acceptance and adoption of AI models in supporting literature assessment.
Conclusion
This study demonstrates the effectiveness of fine-tuning pretrained transformer models for multiclass classification of clinical literature, achieving high performance across most metrics and providing a practical solution for early classification of study types within a knowledge translation workflow. Among the 10 top-performing models, BioBERT variants exhibited superior calibration and consistent results, which could make them well suited for piloting in human-in-the-loop systems to support evidence synthesis workflows. Challenges, such as class imbalance and performance degradation on underrepresented classes (eg, evidence-based guidelines and nonexperimental studies), underscore the need for further exploration of hierarchical classification and strategies to handle rarer classes. Optimal configurations, including lower LRs, midrange BSs, and lower WDs, were critical to achieving robust performance, while CW adjustments proved less beneficial except for recall-focused metrics. These findings contribute to the growing body of research on automated text classification and hold promise for improving the efficiency of clinical literature curation and systematic reviews.
Acknowledgments
We thank the Digital Research Alliance of Canada for the computational resources used in completing this work. FZ received financial support from the Mitacs Business Strategy Internship grant (IT42947) with matching funds from EBSCO. The funding organizations had no involvement in the design, execution, analysis, interpretation, or reporting of this study.
Data Availability
All data and the source code are available upon reasonable request to the first or the corresponding author.
Authors' Contributions
FZ was responsible for conceptualization, data curation, formal analysis, funding acquisition, investigation, methodology, software, validation, visualization, writing—original draft, and writing—review and editing; CL for conceptualization, funding acquisition, investigation, methodology, project administration, resources, supervision, validation, writing—original draft, and writing—review and editing; RP for conceptualization, data curation, formal analysis, investigation, methodology, software, validation, and writing—review and editing; RBH for conceptualization, validation, and writing—review and editing; AI for conceptualization, investigation, resources, and writing—review and editing; AS for validation and writing—review and editing; and MA for conceptualization, investigation, methodology, validation, writing—original draft, and writing—review and editing.
Conflicts of Interest
McMaster University, a public nonprofit academic institution, administers contracts through the Health Information Research Unit under the leadership of AI and RBH. These contracts facilitate collaborations with professional and commercial publishers to obtain newly published studies and systematic reviews, which are critically appraised for research methodology and assessed for clinical relevance as part of the McMaster’s Premium Literature Service (PLUS). CL and RP receive partial remuneration for their involvement in these contracts, while RBH is compensated for supervisory responsibilities and receives royalties. AS, FZ, and MA have no affiliations with PLUS.
Evaluation metric, model training environment, best-performing models, performance of top models, aggregated model performance, confusion matrices ,and calibration plots.
DOC File , 9677 KBMacro averages of evaluation metrics on the PLUS-validate set by model configuration parameters and performance of the top models on AUROC, F1-score, and MCC. AUROC: area under the receiver operating characteristic curve; MCC: Matthew’s correlation coefficient; PLUS: Premium Literature Service.
DOCX File , 24 KBReferences
- MEDLINE PubMed production statistics. National Library of Medicine. Nov 26, 2018. URL: https://www.nlm.nih.gov/bsd/medline_pubmed_production_stats.html [accessed 2025-01-20]
- Densen P. Challenges and opportunities facing medical education. Trans Am Clin Climatol Assoc. 2011;122:48-58. [FREE Full text] [Medline]
- Taha K, Yoo P, Yeun C, Taha A. Text classification: a review, empirical, and experimental evaluation. arXiv. Preprint posted online on 11 Jan, 2024. [FREE Full text] [CrossRef]
- Wan Z. Text classification: a perspective of deep learning methods. arXiv. Preprint posted online on 24 Sep, 2023. [FREE Full text] [CrossRef]
- Kumar S, Roy PP, Dogra DP, Kim B-G. A comprehensive review on sentiment analysis: tasks, approaches and applications. arXiv. Preprint posted online 19 Nov 2023. [FREE Full text] [CrossRef]
- AbdulNabi I, Yaseen Q. Spam email detection using deep learning techniques. Procedia Comput Sci. 2021;184:853-858. [FREE Full text] [CrossRef]
- Markov IL, Liu J, Vagner A. Regular expressions for fast-response COVID-19 text classification. arXiv. Preprint posted online on 18 Feb 2021. [FREE Full text] [CrossRef]
- Kowsari K, Meimandi K, Heidarysafa M, Mendu S, Barnes L, Brown D. Text classification algorithms: a survey. arXiv. Preprint posted online on 17 Apr 2019. [FREE Full text] [CrossRef]
- Michael LI, Donohue J, Davis J, Lee D, Servant F. Regexes are hard: decision-making, difficulties, and risks in programming regular expressions. arXiv. Preprint posted online on 5 Mar 2023. [FREE Full text] [CrossRef]
- Li X-L, Liu B. Rule-based classification. WayBack Machine. URL: https://citeseerx.ist.psu.edu/document?repid=rep1&type=pdf&doi=8557fde1e865f0dcc209468ccaf94f12a04b7835 [accessed 2026-04-09]
- Li Q, Peng H, Li J, Xia C, Yang R, Sun L, et al. A survey on text classification: from shallow to deep learning. arXiv. Preprint posted online on 2 Aug 2020. [FREE Full text] [CrossRef]
- Yin X-C, Yang C, Pei W-Y, Hao H-W. Shallow classification or deep learning: an experimental study. 2014. Presented at: 2014 22nd International Conference on Pattern Recognition; August 24-28, 2014:1904-1909; Stockholm, Sweden. [CrossRef]
- Oleynik M, Kugic A, Kasáč Z, Kreuzthaler M. Evaluating shallow and deep learning strategies for the 2018 n2c2 shared task on clinical text classification. J Am Med Inform Assoc. Nov 01, 2019;26(11):1247-1254. [FREE Full text] [CrossRef] [Medline]
- Xu Y, Zhou Y, Sekula P, Ding L. Machine learning in construction: from shallow to deep learning. Dev Built Environ. May 2021;6:100045. [FREE Full text] [CrossRef]
- Shahid F, Zameer A, Muneeb M. Predictions for COVID-19 with deep learning models of LSTM, GRU and Bi-LSTM. Chaos Solitons Fractals. Nov 2020;140:110212. [FREE Full text] [CrossRef] [Medline]
- Guo L, Zhang D, Wang L, Wang H, Cui B. CRAN: a hybrid CNN-RNN attention-based model for text classification. In: Conceptual Modeling. Cham. Springer; Sep 26, 2018:571-585.
- Zhou Y. A review of text classification based on deep learning. 2020. Presented at: 2020 3rd International Conference on Geoinformatics and Data Analysis; April 15-17, 2020:132-136; Marseille France. [CrossRef]
- Noh S-H. Analysis of gradient vanishing of RNNs and performance comparison. Information. Oct 25, 2021;12(11):442. [FREE Full text] [CrossRef]
- Wu X, Xiang B, Lu H, Li C, Huang X, Huang W. Optimizing recurrent neural networks: a study on gradient normalization of weights for enhanced training efficiency. Appl Sci. Jul 27, 2024;14(15):6578. [FREE Full text] [CrossRef]
- Cui Z, Ke R, Pu Z, Wang Y. Deep bidirectional and unidirectional LSTM recurrent neural network for network-wide traffic speed prediction. arXiv. Preprint posted online on 7 Jan 2018. [FREE Full text] [CrossRef]
- Liu S, Ni’mah I, Menkovski V, Mocanu DC, Pechenizkiy M. Efficient and effective training of sparse recurrent neural networks. Neural Comput Appl. Jan 26, 2021;33(15):9625-9636. [FREE Full text] [CrossRef]
- Pu Q, Xi Z, Yin S, Zhao Z, Zhao L. Advantages of transformer and its application for medical image segmentation: a survey. Biomed Eng Online. Feb 03, 2024;23(1):14. [FREE Full text] [CrossRef] [Medline]
- Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, et al. Attention is all you need. arXiv. Preprint posted online on 12 Jun 2017. [FREE Full text] [CrossRef]
- Zhang H, Shafiq MO. Survey of transformers and towards ensemble learning using transformers for natural language processing. J Big Data. 2024;11(1):25. [FREE Full text] [CrossRef] [Medline]
- Hernández A, Amigó JM. Attention mechanisms and their applications to complex systems. Entropy (Basel). Feb 26, 2021;23(3):283. [FREE Full text] [CrossRef] [Medline]
- Lee J, Yoon W, Kim S, Kim D, Kim S, So CH, et al. BioBERT: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics. Feb 15, 2020;36(4):1234-1240. [FREE Full text] [CrossRef] [Medline]
- Beltagy I, Lo K, Cohan A. SciBERT: a pretrained language model for scientific text. arXiv. Preprint posted online on 26 Mar 2019. [FREE Full text] [CrossRef]
- Yasunaga M, Leskovec J, Liang P. LinkBERT: pretraining language models with document links. arXiv. Preprint posted online on 29 Mar 2022. [FREE Full text] [CrossRef]
- Hosna A, Merry E, Gyalmo J, Alom Z, Aung Z, Azim MA. Transfer learning: a friendly introduction. J Big Data. Oct 22, 2022;9(1):102. [FREE Full text] [CrossRef] [Medline]
- Irwin AN, Rackham D. Comparison of the time-to-indexing in PubMed between biomedical journals according to impact factor, discipline, and focus. Res Social Adm Pharm. 2017;13(2):389-393. [FREE Full text] [CrossRef] [Medline]
- Dhammi IK, Kumar S. Medical subject headings (MeSH) terms. Indian J Orthop. Sep 2014;48(5):443-444. [FREE Full text] [CrossRef] [Medline]
- MEDLINE 2022 initiative: transition to automated indexing. National Library of Medicine. Dec 1, 2021. URL: https://www.nlm.nih.gov/pubs/techbull/nd21/nd21_medline_2022.html [accessed 2024-10-23]
- MTIX: the next-generation algorithm for automated indexing of MEDLINE. National Library of Medicine. Apr 29, 2024. URL: https://www.nlm.nih.gov/pubs/techbull/ma24/ma24_mtix.html [accessed 2024-10-23]
- Thushari P, Niazi S, Meena S. Transfer learning approach to multilabel biomedical literature classification using transformer models. 2023. Presented at: 2023 IEEE 8th International Conference for Convergence in Technology (I2CT); April 7-9, 2023:1-6; Lonavla, India. [CrossRef]
- Chen Q, Du J, Allot A, Lu Z. LitMC-BERT: Transformer-Based Multi-Label Classification of Biomedical Literature With An Application on COVID-19 Literature Curation. IEEE/ACM Trans Comput Biol Bioinform. 2022;19(5):2584-2595. [FREE Full text] [CrossRef] [Medline]
- Verma S, Sharan A, Malik N. Efficient classification of hallmark of cancer using embedding-based support vector machine for multilabel text. New Gener Comput. Mar 12, 2024;42(4):685-714. [FREE Full text] [CrossRef]
- Liu S, Tang H, Liu H, Wang J. Multi-label learning for the diagnosis of cancer and identification of novel biomarkers with High-throughput omics. Curr Bioinform. Feb 2021;16(2):261-273. [FREE Full text] [CrossRef]
- Baker S, Silins I, Guo Y, Ali I, Högberg J, Stenius U, et al. Automatic semantic classification of scientific literature according to the hallmarks of cancer. Bioinformatics. Feb 01, 2016;32(3):432-440. [FREE Full text] [CrossRef] [Medline]
- Kilicoglu H, Demner-Fushman D, Rindflesch TC, Wilczynski NL, Haynes RB. Towards automatic recognition of scientifically rigorous clinical research evidence. J Am Med Inform Assoc. 2009;16(1):25-31. [FREE Full text] [CrossRef] [Medline]
- Aphinyanaphongs Y, Tsamardinos I, Statnikov A, Hardin D, Aliferis CF. Text categorization models for high-quality article retrieval in internal medicine. J Am Med Inform Assoc. 2005;12(2):207-216. [FREE Full text] [CrossRef] [Medline]
- Lokker C, Abdelkader W, Bagheri E, Parrish R, Cotoi C, Navarro T, et al. Boosting efficiency in a clinical literature surveillance system with LightGBM. PLOS Digit Health. Sep 2024;3(9):e0000299. [FREE Full text] [CrossRef] [Medline]
- Lokker C, Bagheri E, Abdelkader W, Parrish R, Afzal M, Navarro T, et al. Deep learning to refine the identification of high-quality clinical research articles from the biomedical literature: performance evaluation. J Biomed Inform. Jun 2023;142:104384. [FREE Full text] [CrossRef] [Medline]
- Marshall IJ, Kuiper J, Wallace BC. RobotReviewer: evaluation of a system for automatically assessing bias in clinical trials. J Am Med Inform Assoc. Jan 2016;23(1):193-201. [FREE Full text] [CrossRef] [Medline]
- van den Bulk LM, Bouzembrak Y, Gavai A, Liu N, van den Heuvel LJ, Marvin HJ. Automatic classification of literature in systematic reviews on food safety using machine learning. Curr Res Food Sci. 2022;5:84-95. [FREE Full text] [CrossRef] [Medline]
- Lange T, Schwarzer G, Datzmann T, Binder H. Machine learning for identifying relevant publications in updates of systematic reviews of diagnostic test studies. Res Synth Methods. Jul 2021;12(4):506-515. [FREE Full text] [CrossRef] [Medline]
- Shekelle PG, Shetty K, Newberry S, Maglione M, Motala A. Machine learning versus standard techniques for updating searches for systematic reviews: a diagnostic accuracy study. Ann Intern Med. Aug 01, 2017;167(3):213-215. [FREE Full text] [CrossRef] [Medline]
- Chernikova O, Stadler M, Melev I, Fischer F. Using machine learning for continuous updating of meta-analysis in educational context. Comput Human Behav. Jul 2024;156:108215. [FREE Full text] [CrossRef]
- Qin X, Liu J, Wang Y, Liu Y, Deng K, Ma Y, et al. Natural language processing was effective in assisting rapid title and abstract screening when updating systematic reviews. J Clin Epidemiol. May 2021;133:121-129. [FREE Full text] [CrossRef] [Medline]
- Aum S, Choe S. srBERT: automatic article classification model for systematic review using BERT. Syst Rev. Oct 30, 2021;10(1):285. [FREE Full text] [CrossRef] [Medline]
- Better systematic review management. Covidence. 2020. URL: https://www.covidence.org/ [accessed 2024-12-03]
- Intelligent systematic review. Rayyan. 2021. URL: https://www.rayyan.ai/ [accessed 2024-12-03]
- Liu Y, Bi J, Fan Z. A method for multi-class sentiment classification based on an improved one-vs-one (OVO) strategy and the support vector machine (SVM) algorithm. Inf Sci. Jul 2017;394-395:38-52. [FREE Full text] [CrossRef]
- Song Y, Zhang J, Yan H, Li Q. Multi-class imbalanced learning with one-versus-one decomposition: an empirical study. In: Lecture Notes in Computer Science. Cham. Springer; Sep 13, 2018:617-628.
- Wolf DA, Galin RR. Performance comparative analysis of OvA, AvA, and OvO algorithms in multi-class classification. 2024. Presented at: 2024 17th International Conference on Management of Large-Scale System Development (MLSD); September 24-26, 2024:1-5; Moscow, Russian Federation. [CrossRef]
- Hong J, Cho S. A probabilistic multi-class strategy of one-vs.-rest support vector machines for cancer classification. Neurocomputing. Oct 2008;71(16-18):3275-3281. [FREE Full text] [CrossRef]
- Ou G, Murphey YL. Multi-class pattern classification using neural networks. Pattern Recognit. Jan 2007;40(1):4-18. [FREE Full text] [CrossRef]
- Afzal M, Hussain J, Abbas A, Hussain M, Attique M, Lee S. Transformer-based active learning for multi-class text annotation and classification. Digit Health. 2024;10:20552076241287357. [FREE Full text] [CrossRef] [Medline]
- Rabby G, Berka P. Multi-class classification of COVID-19 documents using machine learning algorithms. J Intell Inf Syst. Nov 29, 2023;60(2):571-591. [FREE Full text] [CrossRef] [Medline]
- Wilczynski NL, Morgan D, Haynes RB, Hedges Team. An overview of the design and methods for retrieving high-quality studies for clinical care. BMC Med Inform Decis Mak. Jun 21, 2005;5:20. [FREE Full text] [CrossRef] [Medline]
- McMaster Health Knowledge Refinery (HKR). McMaster University. URL: https://hiruweb.mcmaster.ca/hkr/hedges/ [accessed 2026-04-09]
- Haynes RB, Holland J, Cotoi C, McKinlay RJ, Wilczynski NL, Walters LA, et al. McMaster PLUS: a cluster randomized clinical trial of an intervention to accelerate clinical use of evidence-based information from digital libraries. J Am Med Inform Assoc. Nov 2006;13(6):593-600. [FREE Full text] [CrossRef] [Medline]
- Jin D, Szolovits P. Advancing PICO element detection in biomedical text via deep neural networks. Bioinformatics. Jun 01, 2020;36(12):3856-3862. [FREE Full text] [CrossRef] [Medline]
- Naseem U, Khushi M, Reddy V, Rajendran S, Razzak I, Kim J. BioALBERT: a simple and effective pre-trained language model for biomedical named entity recognition. arXiv. Preprint posted online on 19 Sep 2020. [FREE Full text] [CrossRef]
- Mutinda FW, Liew K, Yada S, Wakamiya S, Aramaki E. Automatic data extraction to support meta-analysis statistical analysis: a case study on breast cancer. BMC Med Inform Decis Mak. Jun 18, 2022;22(1):158. [FREE Full text] [CrossRef] [Medline]
- Moradi M, Dorffner G, Samwald M. Deep contextualized embeddings for quantifying the informative content in biomedical text summarization. Comput Methods Programs Biomed. Feb 2020;184:105117. [FREE Full text] [CrossRef] [Medline]
- Ding H, Luo X. Attention-based unsupervised keyphrase extraction and phrase graph for COVID-19 medical literature retrieval. ACM Trans Comput Healthcare. Oct 15, 2021;3(1):1-16. [FREE Full text] [CrossRef]
- Kanakarajan KR, Kundumani B, Sankarasubbu M. BioELECTRA: pretrained biomedical text encoder using discriminators. In: Demner-Fushman D, Cohen KB, Ananiadou S, Tsujii J, editors. Proceedings of the 20th Workshop on Biomedical Language Processing. Stroudsburg, PA. Association for Computational Linguistics; Jun 2021:143-154.
- BLURB leaderboard. Microsoft. URL: https://microsoft.github.io/BLURB/leaderboard.html [accessed 2026-04-09]
- Loshchilov I, Hutter F. Decoupled weight decay regularization. arXiv. Preprint posted online on 14 Nov 2017. [FREE Full text] [CrossRef]
- Gu Y, Tinn R, Cheng H, Lucas M, Usuyama N, Liu X, et al. Domain-specific language model pretraining for biomedical natural language processing. ACM Trans Comput Healthcare. Oct 15, 2021;3(1):1-23. [FREE Full text] [CrossRef]
- Grandini M, Bagli E, Visani G. Metrics for multi-class classification: an overview. arXiv. Preprint posted online on 13 Aug 2020. [FREE Full text] [CrossRef]
- Rutter CM. Bootstrap estimation of diagnostic accuracy with patient-clustered data. Acad Radiol. Jun 2000;7(6):413-419. [FREE Full text] [CrossRef] [Medline]
- Heßler N, Rottmann M, Ziegler A. Empirical analysis of the text structure of original research articles in medical journals. PLoS One. 2020;15(10):e0240288. [FREE Full text] [CrossRef] [Medline]
- Markey N, Howitt B, El-Mansouri I, Schwartzenberg C, Kotova O, Meier C. Clinical trials are becoming more complex: a machine learning analysis of data from over 16,000 trials. Sci Rep. Feb 12, 2024;14(1):3514. [FREE Full text] [CrossRef] [Medline]
- Sahiner B, Chen W, Samala RK, Petrick N. Data drift in medical machine learning: implications and potential remedies. Br J Radiol. Oct 2023;96(1150):20220878. [FREE Full text] [CrossRef] [Medline]
- Kumar V, Lalotra GS, Sasikala P, Rajput DS, Kaluri R, Lakshmanna K, et al. Addressing binary classification over class imbalanced clinical datasets using computationally intelligent techniques. Healthcare (Basel). Jul 13, 2022;10(7):1293. [FREE Full text] [CrossRef] [Medline]
- Welvaars K, Oosterhoff JHF, van den Bekerom MPJ, Doornberg JN, van Haarst EP, OLVG Urology Consortium‚ the Machine Learning Consortium. Implications of resampling data to address the class imbalance problem (IRCIP): an evaluation of impact on performance between classification algorithms in medical data. JAMIA Open. Jul 2023;6(2):ooad033. [FREE Full text] [CrossRef] [Medline]
- van den Goorbergh R, van Smeden M, Timmerman D, Van Calster B. The harm of class imbalance corrections for risk prediction models: illustration and simulation using logistic regression. J Am Med Inform Assoc. Aug 16, 2022;29(9):1525-1534. [FREE Full text] [CrossRef] [Medline]
- Caplin A, Martin D, Marx P. Calibrating for class weights by modeling machine learning. arXiv. Preprint posted online on 10 May 2022. [FREE Full text] [CrossRef]
- Piccininni M, Wechsung M, Van Calster B, Rohmann JL, Konigorski S, van Smeden M. Understanding random resampling techniques for class imbalance correction and their consequences on calibration and discrimination of clinical risk prediction models. J Biomed Inform. Jul 2024;155:104666. [FREE Full text] [CrossRef] [Medline]
- Sun C, Qiu X, Xu Y, Huang X. How to fine-tune BERT for text classification? In: Lecture Notes in Computer Science. Cham. Springer; 2019:194-206.
- Nagatsuka K, Broni-Bediako C, Atsumi M. Pre-training a BERT with curriculum learning by increasing block-size of input text. In: Mitkov R, Angelova G, editors. Proceedings of the Conference Recent Advances in Natural Language Processing - Deep Learning for Natural Language Processing Methods and Applications. Shoumen, Bulgaria. INCOMA; 2021:989-996.
- Huo H, Iwaihara M. Utilizing BERT pretrained models with various fine-tune methods for subjectivity detection. In: Lecture Notes in Computer Science. Cham. Springer; 2020:270-284.
- Li F, Jin Y, Liu W, Rawat BPS, Cai P, Yu H. Fine-tuning bidirectional encoder representations from transformers (BERT)-based models on large-scale electronic health record notes: an empirical study. JMIR Med Inform. Sep 12, 2019;7(3):e14830. [FREE Full text] [CrossRef] [Medline]
- Zhao X, Jiang H, Yin J, Liu H, Zhu R, Mei S, et al. Changing trends in clinical research literature on PubMed database from 1991 to 2020. Eur J Med Res. Jun 20, 2022;27(1):95. [FREE Full text] [CrossRef] [Medline]
- Zliobaite I, Bifet A, Pfahringer B, Holmes G. Active learning with drifting streaming data. IEEE Trans Neural Netw Learn Syst. Jan 2014;25(1):27-39. [FREE Full text] [CrossRef] [Medline]
- Liu S, Xue S, Wu J, Zhou C, Yang J, Li Z, et al. Online active learning for drifting data streams. IEEE Trans Neural Netw Learn Syst. Jan 2023;34(1):186-200. [FREE Full text] [CrossRef] [Medline]
- Health EvidenceTM. National Collaborating Centre for Methods and Tools. URL: https://www.healthevidence.org/about-us.aspx [accessed 2025-08-14]
- Rogers K, Miller A, Girgis A, Clark EC, Neil-Sztramko SE, Dobbins M. Leveraging AI to optimize maintenance of health evidence and offer a one-stop shop for quality-appraised evidence syntheses on the effectiveness of public health interventions: quality improvement project. J Med Internet Res. Jul 29, 2025;27:e69700. [FREE Full text] [CrossRef] [Medline]
- Cinteza M. Artificial intelligence and copyright. Maedica (Bucur). Mar 2025;20(1):1-6. [FREE Full text] [CrossRef] [Medline]
- Raja H, Munawar A, Mylonas N, Delsoz M, Madadi Y, Elahi M, et al. Automated category and trend analysis of scientific articles on ophthalmology using large language models: development and usability study. JMIR Form Res. Mar 22, 2024;8:e52462. [FREE Full text] [CrossRef] [Medline]
- Joshi S, Abdelfattah E. Multi-class text classification using machine learning models for online drug reviews. 2021. Presented at: 2021 IEEE World AI IoT Congress (AIIoT); May 10-13, 2021; Seattle, WA. [CrossRef]
- Sundararajan M, Taly A, Yan Q. Axiomatic attribution for deep networks. arXiv. Preprint posted online on 4 Mar 2017. [FREE Full text] [CrossRef]
- Zhou F. Generative large language models for transparent artificial intelligence in clinical research: enhancing interpretability through appraisal and explanation. McMaster University. 2025. URL: http://hdl.handle.net/11375/31871 [accessed 2025-09-21]
- Lundberg S, Lee S-I. A unified approach to interpreting model predictions. arXiv. Preprint posted online on 22 May 2017 . [FREE Full text] [CrossRef]
Abbreviations
| AI: artificial intelligence |
| AP: average precision |
| AUROC: area under the receiver operating characteristic curve |
| BART: Bidirectional and Auto-Regressive Transformers |
| BERT: Bidirectional Encoder Representations from Transformers |
| BioBERT: Biomedical Bidirectional Encoder Representations from Transformers (fine-tuned on biomedical text) |
| BioELECTRA: Biomedical Efficiently Learning an Encoder that Classifies Token Replacements Accurately |
| BioLinkBERT: Biomedical Document Link Bidirectional Encoder Representations from Transformers |
| BiomedBERT: Biomedical Bidirectional Encoder Representations from Transformers (trained entirely on biomedical text) |
| BS: batch size |
| CW: class weight |
| GPU: graphics processing unit |
| GRU: Gated recurrent unit |
| HIRU: Health Information Research Unit |
| LR: learning rate |
| MCC: Matthew’s correlation coefficient |
| MeSH: Medical Subject Headings |
| ML: machine learning |
| OvO: one versus one |
| OvR: one versus rest |
| PLUS: Premium Literature Service |
| SciBERT: Scientific Bidirectional Encoder Representations from Transformers |
| SL: shallow learning |
| SOAP: Subjective Objective Assessment Plan |
| TF-IDF: term frequency–inverse document frequency |
| UMLS: Unified Medical Language System |
| WD: weight decay |
| WR: warmup ratio |
Edited by K El Emam; submitted 12.May.2025; peer-reviewed by H Jelodar, S Mohanadas, H Maheshwari, C Obianyo; comments to author 08.Aug.2025; revised version received 21.Sep.2025; accepted 17.Mar.2026; published 29.Apr.2026.
Copyright©Fangwen Zhou, Cynthia Lokker, Rick Parrish, R Brian Haynes, Alfonso Iorio, Ashirbani Saha, Muhammad Afzal. Originally published in JMIR AI (https://ai.jmir.org), 29.Apr.2026.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR AI, is properly cited. The complete bibliographic information, a link to the original publication on https://www.ai.jmir.org/, as well as this copyright and license information must be included.


